Goto

Collaborating Authors

 graph-based model


Rethinking Node Representation Interpretation through Relation Coherence

Lin, Ying-Chun, Neville, Jennifer, Becker, Cassiano, Metha, Purvanshi, Asghar, Nabiha, Agarwal, Vipul

arXiv.org Artificial Intelligence

Understanding node representations in graph-based models is crucial for uncovering biases ,diagnosing errors, and building trust in model decisions. However, previous work on explainable AI for node representations has primarily emphasized explanations (reasons for model predictions) rather than interpretations (mapping representations to understandable concepts). Furthermore, the limited research that focuses on interpretation lacks validation, and thus the reliability of such methods is unclear. We address this gap by proposing a novel interpretation method-Node Coherence Rate for Representation Interpretation (NCI)-which quantifies how well different node relations are captured in node representations. We also propose a novel method (IME) to evaluate the accuracy of different interpretation methods. Our experimental results demonstrate that NCI reduces the error of the previous best approach by an average of 39%. We then apply NCI to derive insights about the node representations produced by several graph-based methods and assess their quality in unsupervised settings.


A Review of Graph-Powered Data Quality Applications for IoT Monitoring Sensor Networks

Ferrer-Cid, Pau, Barcelo-Ordinas, Jose M., Garcia-Vidal, Jorge

arXiv.org Artificial Intelligence

The use of graph-based models had already been a key element in applications such as route planning (e.g., Dijkstra's algorithm) [3], community detection (e.g., clique percolation or Louvain algorithms) [4] or the analysis of complex networks such as biological and social networks [5, 6]. The representation of manifolds by means of graphs in the field of semi-supervised learning is another example of graph-powered applications [7]. Overall, the GSP framework [8] has enabled the use and development of novel techniques on data residing in graphs, thus emerging as an alternative to classical machine learning techniques that do not make explicit use of data structure. In this way, the graph topology, which represents the relationships between the graph's nodes, is fed to graph-based models that explicitly model the structure of the data [9]. A wide variety of concepts have been applied to signals defined over graphs, such as signal shift, translation, convolution, or filtering [10]. An important concept of the GSP is the notion of signal smoothness, also expressed via the total variation (TV) or the Dirichlet energy, which are quadratic forms and can be used to evaluate how a signal fits a given graph structure or vice-versa [11]. This idea is linked with the Graph Discrete Fourier Transform (GDFT) that makes use of the graph topology to obtain the graph Fourier basis and allow the computation of the transform coefficients of a graph signal and also led to the development of graph filters [12]. In the field of machine learning, graphs have been used as regularizers in optimization problems, e.g., the regularization of neural networks for semi-supervised learning tasks [13].


Chain-of-History Reasoning for Temporal Knowledge Graph Forecasting

Xia, Yuwei, Wang, Ding, Liu, Qiang, Wang, Liang, Wu, Shu, Zhang, Xiaoyu

arXiv.org Artificial Intelligence

Temporal Knowledge Graph (TKG) forecasting aims to predict future facts based on given histories. Most recent graph-based models excel at capturing structural information within TKGs but lack semantic comprehension abilities. Nowadays, with the surge of LLMs, the LLM-based TKG prediction model has emerged. However, the existing LLM-based model exhibits three shortcomings: (1) It only focuses on the first-order history for prediction while ignoring high-order historical information, resulting in the provided information for LLMs being extremely limited. (2) LLMs struggle with optimal reasoning performance under heavy historical information loads. (3) For TKG prediction, the temporal reasoning capability of LLM alone is limited. To address the first two challenges, we propose Chain-of-History (CoH) reasoning which explores high-order histories step-by-step, achieving effective utilization of high-order historical information for LLMs on TKG prediction. To address the third issue, we design CoH as a plug-and-play module to enhance the performance of graph-based models for TKG prediction. Extensive experiments on three datasets and backbones demonstrate the effectiveness of CoH.


Source Code is a Graph, Not a Sequence: A Cross-Lingual Perspective on Code Clone Detection

Rahaman, Mohammed Ataaur, Ive, Julia

arXiv.org Artificial Intelligence

Source code clone detection is the task of finding code fragments that have the same or similar functionality, but may differ in syntax or structure. This task is important for software maintenance, reuse, and quality assurance (Roy et al. 2009). However, code clone detection is challenging, as source code can be written in different languages, domains, and styles. In this paper, we argue that source code is inherently a graph, not a sequence, and that graph-based methods are more suitable for code clone detection than sequence-based methods. We compare the performance of two state-of-the-art models: CodeBERT (Feng et al. 2020), a sequence-based model, and CodeGraph (Yu et al. 2023), a graph-based model, on two benchmark data-sets: BCB (Svajlenko et al. 2014) and PoolC (PoolC no date). We show that CodeGraph outperforms CodeBERT on both data-sets, especially on cross-lingual code clones. To the best of our knowledge, this is the first work to demonstrate the superiority of graph-based methods over sequence-based methods on cross-lingual code clone detection.


Graph vs. Sequence: An Empirical Study on Knowledge Forms for Knowledge-Grounded Dialogue

Yang, Yizhe, Huang, Heyan, Liu, Yihang, Gao, Yang

arXiv.org Artificial Intelligence

Knowledge-grounded dialogue is a task of generating an informative response based on both the dialogue history and external knowledge source. In general, there are two forms of knowledge: manually annotated knowledge graphs and knowledge text from website. From various evaluation viewpoints, each type of knowledge has advantages and downsides. To further distinguish the principles and determinants from the intricate factors, we conduct a thorough experiment and study on the task to answer three essential questions. The questions involve the choice of appropriate knowledge form, the degree of mutual effects between knowledge and the model selection, and the few-shot performance of knowledge. Supported by statistical shreds of evidence, we offer conclusive solutions and sensible suggestions for directions and standards of future research.


Document AI: A Comparative Study of Transformer-Based, Graph-Based Models, and Convolutional Neural Networks For Document Layout Analysis

Kastanas, Sotirios, Tan, Shaomu, He, Yi

arXiv.org Artificial Intelligence

Document AI aims to automatically analyze documents by leveraging natural language processing and computer vision techniques. One of the major tasks of Document AI is document layout analysis, which structures document pages by interpreting the content and spatial relationships of layout, image, and text. This task can be image-centric, wherein the aim is to identify and label various regions such as authors and paragraphs, or text-centric, where the focus is on classifying individual words in a document. Although there are increasingly sophisticated methods for improving layout analysis, doubts remain about the extent to which their findings can be generalized to a broader context. Specifically, prior work developed systems based on very different architectures, such as transformer-based, graph-based, and CNNs. However, no work has mentioned the effectiveness of these models in a comparative analysis. Moreover, while language-independent Document AI models capable of knowledge transfer have been developed, it remains to be investigated to what degree they can effectively transfer knowledge. In this study, we aim to fill these gaps by conducting a comparative evaluation of state-of-the-art models in document layout analysis and investigating the potential of cross-lingual layout analysis by utilizing machine translation techniques.


Predicting COVID-19 pandemic by spatio-temporal graph neural networks: A New Zealand's study

Nguyen, Viet Bach, Hy, Truong Son, Tran-Thanh, Long, Nghiem, Nhung

arXiv.org Artificial Intelligence

Modeling and simulations of pandemic dynamics play an essential role in understanding and addressing the spreading of highly infectious diseases such as COVID-19. In this work, we propose a novel deep learning architecture named Attention-based Multiresolution Graph Neural Networks (ATMGNN) that learns to combine the spatial graph information, i.e. geographical data, with the temporal information, i.e. timeseries data of number of COVID-19 cases, to predict the future dynamics of the pandemic. The key innovation is that our method can capture the multiscale structures of the spatial graph via a learning to cluster algorithm in a data-driven manner. This allows our architecture to learn to pick up either local or global signals of a pandemic, and model both the long-range spatial and temporal dependencies. Importantly, we collected and assembled a new dataset for New Zealand. We established a comprehensive benchmark of statistical methods, temporal architectures, graph neural networks along with our spatio-temporal model. We also incorporated socioeconomic cross-sectional data to further enhance our prediction. Our proposed model have shown highly robust predictions and outperformed all other baselines in various metrics for our new dataset of New Zealand along with existing datasets of England, France, Italy and Spain. For a future work, we plan to extend our work for real-time prediction and global scale.


Spatio-Temporal Attention in Multi-Granular Brain Chronnectomes for Detection of Autism Spectrum Disorder

Orme-Rogers, James, Srivastava, Ajitesh

arXiv.org Artificial Intelligence

The traditional methods for detecting autism spectrum disorder (ASD) are expensive, subjective, and time-consuming, often taking years for a diagnosis, with many children growing well into adolescence and even adulthood before finally confirming the disorder. Recently, graph-based learning techniques have demonstrated impressive results on resting-state functional magnetic resonance imaging (rs-fMRI) data from the Autism Brain Imaging Data Exchange (ABIDE). We introduce IMAGIN, a multI-granular, Multi-Atlas spatio-temporal attention Graph Isomorphism Network, which we use to learn graph representations of dynamic functional brain connectivity (chronnectome), as opposed to static connectivity (connectome). The experimental results demonstrate that IMAGIN achieves a 5-fold cross-validation accuracy of 79.25%, which surpasses the current state-of-the-art by 1.5%. In addition, analysis of the spatial and temporal attention scores provides further validation for the neural basis of autism.


SQuARE: Software for Question Answering Research

#artificialintelligence

Have you ever wanted to try Question Answering (QA) models but felt restrained because you needed to write some code to set them up? Have you ever wanted to compare QA models, but a Jupyter Notebook is too inconvenient to compare them? Have you ever wanted to use explainability methods such as saliency maps to explain the outputs, but you don't even know where to start? We have been there too! That's why we built SQuARE: Software for Question Answering Research!


Parsing Thai Social Data: A New Challenge for Thai NLP

Singkul, Sattaya, Khampingyot, Borirat, Maharattamalai, Nattasit, Taerungruang, Supawat, Chalothorn, Tawunrat

arXiv.org Artificial Intelligence

Dependency parsing (DP) is a task that analyzes text for syntactic structure and relationship between words. DP is widely used to improve natural language processing (NLP) applications in many languages such as English. Previous works on DP are generally applicable to formally written languages. However, they do not apply to informal languages such as the ones used in social networks. Therefore, DP has to be researched and explored with such social network data. In this paper, we explore and identify a DP model that is suitable for Thai social network data. After that, we will identify the appropriate linguistic unit as an input. The result showed that, the transition based model called, improve Elkared dependency parser outperform the others at UAS of 81.42%.